Journal of the College of Physicians and Surgeons Pakistan
ISSN: 1022-386X (PRINT)
ISSN: 1681-7168 (ONLINE)
Affiliations
doi: 10.29271/jcpsp.2025.01.90ABSTRACT
Objective: To establish the construct validity of the Assessment Implementation Measures (AIM) tool to accurately assess faculty perspectives on implemented assessment systems, facilitating the alignment with set standards.
Study Design: Qualitative Study.
Place and Duration of the Study: The study was carried out at RIPHAH International University and data were collected from participants (senior faculty members) involved in teaching and assessment of undergraduates from various medical and dental colleges in Pakistan. Google form questionnaire was distributed using email and WhatsApp. The data were analysed using IBM AMOS SPSS version 24.
Methodology: It was a 30-item tool. The sample size was calculated using 1:10 item-to-participant ratio. Hence data were collected from 313 participants. Confirmatory factor analysis was done to establish construct validity.
Results: The indices of confirmatory factor analysis of the tool showed a discrepancy. To remove this discrepancy, items with weak item-loading values were removed. In the end, the tool was reduced to 13 items belonging to three domains.
Conclusion: The final model was improved by excluding items from the original model. A re-validation study with a careful selection of experienced participants from various institutional backgrounds having baseline knowledge of medical education is suggested to improve the results.
Key Words: AIM Tool, Confirmatory factor analysis, Medical education, Assessment.
INTRODUCTION
Assessment is crucial in medical education, as it evaluates student capabilities and fosters learning through continuous feedback.1 Formative assessment is gaining importance over summative assessment due to its ongoing nature. A curriculum must clearly define assessment policies, including fail/pass criteria, promotion/demotion cases, and external scrutiny, to ensure effective preparation by stakeholders.2 Additionally, balanced weightage across knowledge, skill, and attitude domains is essential for a comprehensive assessment strategy. Without these elements, the curriculum fails to serve its intended purpose.3
Effective student evaluation requires faculty orientation on institutional assessment policies and procedures.
Faculty development programmes ought to instruct instructors in the use of assessment instruments, the creation of challenging multiple-choice questions, the creation of question banks, and the analysis of post-exam items.4 Though many are ignorant of important principles such as the pass/fail criteria and exam retakes, teachers' opinions on assessment are quite important. Their selection of assessment instruments is impacted by this ignorance or ambiguous communication, which may jeopardise the accuracy of the evaluation of students' knowledge and proficiency.
There has been no significant study assessing faculty perceptions of assessment measures in undergraduate medical curricula. The only identified tool for this purpose is the Assessment Implementation Measures (AIM) tool, which evaluates perceptions across four domains: Assessment policies (AP), assessment quality measures (AQM), purpose of assessment (PA), and assessment methods (AM).5 These domains include 8, 9, 5, and 8 items, respectively, making up a 30-item tool. Responses are measured on a 5-point Likert scale from strongly agree (score 4) to strongly disagree (score 0). The maximum score of 120 indicates ideal assessment quality as perceived by the respondent, supporting institutional self-evaluation and enhancement of assessment practices.
The important factor analysis required for tool validation was not carried out during the AIM tool's initial piloting. To make sure the tool measures what it is supposed to and yields correct findings, validation is crucial. In the absence of this, the gathered data may be skewed and inaccurate.6 By using this method to better understand faculty perceptions, assessment quality can be improved and problems with assessment design can be found, facilitating self-evaluation. By establishing the construct validity of the AIM tool using factor analysis, this study aims to close the gap that exists between established assessment standards and real-world undergraduate medical programme practices.7
The purpose of this study was to establish construct validity by doing factor analysis of the AIM tool which would enable us to bridge the gap between set standards for assessment and implemented practices in medical undergraduate programmes in each institution. Through the AIM tool, which was developed to evaluate teachers’ perspectives, an insight about what are the various factors that contribute to the successful implementation of assessment procedures as well as the issues that hinder the process.
METHODOLOGY
The study commenced after approval from the Institutional Review Board (IRB)under Ref. No. Riphah/IIMC/IRC/22/2010. The study was completed in 7 months. The questionnaire was circulated to the participants through email and WhatsApp using Google forms. Detailed description of this study along with informed consent form was furnished to every participant in the start of the Google form. Anonymity and confidentiality of data were maintained at all levels of this research project. As this was a tool-validation study, written permission was obtained by the author to use and validate the tool. Participants’ identity was kept anonymous and only the master researcher had access to the data. It is a tool-validation study, based on the guidelines laid by Association for Medical Education in Europe (AMEE) guide no. 87. Tool validation involves step 7 of the AMEE guide.
The non-probability purposive sampling technique was used. In this type of sampling technique, the researcher himself selects the participants because of their location, position, job specification, education, or socioeconomic status etc. (httpss://www. cambridge.org/core/journals/prehospital-and-disaster-medi cine/article/population-research-convenience-samplingstrate gies/B0D519269C76DB5BFFBFB84ED7031267). For this study, all the participants were faculty members involved in undergraduate medical and dental teaching, working in various private and government colleges across Pakistan. Participants comprised of faculty members from basic medical and dental sciences as well as clinical sciences. However, junior lecturers and non-teaching clinical faculty were excluded from the study. The suggested sample size for factor analyses ranges from 50 to over 1,000 samples. Additionally, the recommended item-to-response ratio falls between 1:3 and 1:20. There were total 30 items and using an average item-to- response ratio of 1:10, the sample size of 300 was calculated.8 As this tool is comprised of 30 items, so the sample size was calculated to be 300. Data analysis involved factor analysis of the items which was done using IBM AMOS (analysis of a moment structures) SPSS version 24. AMOS is indeed a plug-in for SPSS that is used for structural equation modelling (SEM), available from: (httpss://www.taylorfrancis.com/books/mono/ 10.4324/9781003018414/applied-structural-equation-modeling- using-amos-joel-collier).
This software is used to analyse data by doing factor analysis and SEM.9 A confirmatory factor analysis (CFA) with maximum likelihood method was conducted. In evaluation of confirmatory factor analysis, absolute and relative fit indices are used. Absolute fit indices determine how well the prior model fits or reproduces data. The principal component analysis (PCA) is used to explore underlying structures. (httpss://onlinelibrary.wiley. com/doi/abs/10.1002/9781119111931.ch158).
They include chi-square test, root mean-square error of approximation (RMSEA), goodness of fit index (GFI), adjusted goodness of fit index (AGFI), root mean-square residual (RMR), and standardised root mean-square residual (SRMR). The data analysis process using AMOS involved initial data reduction. It was performed in SPSS by putting in all variables into the data reduction option. Principal components analysis with Varimax rotation was applied to explore underlying structures and more interpretable factor structure in PCA. The results obtained from the rotation matrix were transferred to the pattern matrix within the AMOS model builder interface. Refinement of the path diagram was carried out by incorporating guidance from model fit measures, enhancing the clarity and accuracy of the structural model. Counts and percentages were used to express categorical variables.
RESULTS
Of all the 313 faculty members who submitted completed questionnaires, 159 (50.8%) were males and 154 (49.2%) were females. One hundred and forty-five (46.3%) faculty members had more than 11 years of teaching experience, whereas 93 (29.7%) had 8-11 years, 32 (10.2%) had 4-7 years, and 41 (13.1%) had up to 3 years of teaching experience. Of all the respondents, 59 (18.9%) were senior registrars, 132 (42.2%) were assistant professors, 67 (21.4%) were associate professors, and 53 (17.0%) were professors.
Table I shows indices such as minimum discrepancy function by degrees of freedom divided (CMIN/DF), comparative fit index (CFI), SRMR, root mean square error of approximation (RMSEA), and PClose showed acceptable values for modified AIM tool. The internal consistency of the tool was estimated through Cronbach’s Alpha, and it came out to be 0.93 which is in acceptable limit (0.1-1). Figure 1 shows factor analysis of original AIM tool with item loading in 4 domains i.e., (A) assessment methods, (B) purpose of assessment, (C) assessment policies, and (D) assessment quality measures.
Table I: Model fit measures of AIM tool.
Measure |
Estimate |
Threshold |
Interpretation |
CMIN/DF |
2.463 |
Between 1 and 3 |
Excellent |
CFI |
0.844 |
>0.95 |
Terrible |
SRMR |
0.075 |
<0.08 |
Excellent |
RMSEA |
0.069 |
<0.06 |
Acceptable |
PClose |
0.06 |
>0.05 |
Acceptable |
Table II: Validity measures of AIM tool.
Measure |
CR |
AVE |
MSV |
MaxR (H) |
A |
B |
C |
D |
A |
0.845 |
0.382 |
0.580 |
0.854 |
0.618 |
|
|
|
B |
0.859 |
0.360 |
0.582 |
0.869 |
0.763*** |
0.601 |
|
|
C |
0.836 |
0.470 |
0.339 |
0.875 |
0.582*** |
0.576*** |
0.685 |
|
D |
0.808 |
0.530 |
0.339 |
0.885 |
0.583*** |
0.567*** |
0.410*** |
0.728 |
Table III: Model fit measures of assessment methods.
Measure |
Estimate |
Threshold |
Interpretation |
CMIN/DF |
2.394 |
Between 1 and 3 |
Excellent |
CFI |
0.925 |
>0.95 |
Excellent |
SRMR |
0.058 |
<0.08 |
Excellent |
RMSEA |
0.067 |
<0.06 |
Acceptable |
PClose |
0.111 |
>0.05 |
Excellent |
Figure 1: Factor analysis of original AIM tool with item loading in four domains.
Composite reliability (CR) and average variance extracted (AVE) which are measures of internal reliability and consistency are presented in Table II. AVE of each domain or construct must be at least 0.5. The value of AVE in original tool for domain A, B, and C showed values less than 0.5. The tool was modified using AMOS software and items with low item-loading were removed and the AVE was improved.
Figure 2: The factor analysis of modified AIM tool.
As a result, the domain of assessment methods was removed. Items with low item-loading were removed from the remaining three domains. Modified tool with good AVE is shown in Table II and model fit measures is shown in Table III. The factor analysis of the modified AIM tool is shown in Figure 2.
DISCUSSION
The main objective of the study was to validate AIM tool that can be used to assess the perspectives of teachers on prevailing assessment system in their institutions. The results showed that out of 30 items in four domains, only 13 items in three domains could be validated. AMOS software designed for SEM, path analysis, and, CFA, provided proof of item-loading into each domain and eliminated the items with poor item-loading. In CFA, the researcher specifies the proposed relationships among latent factors and their observed indicators using a SEM framework. (httpss://www.tand fonline.com/doi/abs/10.1080/00273171.2019.1602503). It defines the relationship between latent factors (items) and observed variables (indicator loadings), as well as the covariance among latent factors. Fit indices are used to evaluate how well the hypothesised model fits the observed data. Common fit indices include the Chi-square statistic, CFI, and SRMR.10 Lower values of RMSEA and SRMR and higher values of CFI and TLI, indicate better fit model. If the initial model fit is unsatisfactory, modifications may be necessary to improve fit. This could involve adding or removing paths, allowing correlated errors between indicators, or re-specifying the model based on theoretical considerations. Here, the items with lower factor loading were removed to improve the values.11
This tool can be used by teachers, college administration, and examination authority for assessing the quality of assessments and comparing it with the international standards. (httpss:// dro.deakin.edu.au/articles/journal_contribution/Challenges_ in_reforming_higher_education_assessment_a_perspective_ from_afar/20691955/1). There are many studies reported in literature that measure medical undergraduate students’ perceptions regarding assessment in various aspects such as tools of assessment, grading system, and level of difficulty (httpss://asmepublications.onlinelibrary.wiley.com/doi/abs/10. 1046/j.1365-2923.2002.01291.x) but the perceptions of teachers related to various aspects of assessment are not explored in depth. Sajjad et al. constructed assessment implementation measures tool for assessing faculty knowledge and perspectives about assessment practices in their institutions compared to national and international standards.5 This tool was developed using a mixed-method approach. Literature review was done to make a primary questionnaire and was given to 10 medical educationists for three rounds modified-delphi technique. Panel agreement of ≥75% was considered for inclusion of items. Cognitive pre-testing and later piloting was done by randomly selecting 30 faculty members. The value of Cronbach’s alpha was estimated to be 0.93 which is a measure of internal consis-tency and reliability of the tool (httpss://www.ncbi.nlm.nih. gov/pmc/articles/PMC4205511/).
Internal consistency pertains to how closely all items within a test align with the same underlying concept or construct, thus reflecting the inter-relatedness of the items within the test or tool.12 The original AIM tool covers assessment policies, assessment purpose, assessment methods, and assessment quality measures. The domain of assessment methods was removed during factor analysis which led to an incidental finding that the responses to various items in that domain needed knowledge of health professionals education. Lack of knowledge of faculty members who did not have any medical education background led to faulty responses in that domain that led to lower values of covariance and hence their item loading was also not up to the mark. (httpss://www.science direct.com/science/article/abs/pii/S000296102100341X).
The main strength of this study was to establish construct validity and reliability of AIM tool on the basis of an empirical approach with SEM, which is a popular technique for establishing the validity and reliability of the instrument. Another strength of this study was to engage faculty members from varying backgrounds and experiences, i.e., from lecturers, assistant professors, associate professors, and professors to heads of institutions. However, this study is not without limitations. The sample of the study was 311 cases collected from different designations which had different exposure about current assessment practices and looked at assessment with different lenses. A more in-depth study with a greater number of faculty members, especially with health professions education background should be done for better results.
CONCLUSION
Knowing the faculty's perception about standards for the assessment and their state of implementation in the undergraduate medical curriculum helps the institutions to improve their teaching and assessment strategies. Factor-loading indicated the need to remove many items whose model-fit values were not up to the mark. Hence the final modified model with all best indices values showed only three dimensions i.e., assessment policies (AP) with five items, Assessment quality measures (AQM) with three items, and purpose of assessment (PA) with three items. In this model, assessment methods (AM) were removed.
ETHICAL APPROVAL:
The study commenced after approval from the Institutional Review Board (IRB) under Ref. No. Riphah/IIMC/IRC/22/2010.
PARTICIPANTS’ CONSENT:
Informed consent was taken from all the participants included in the study.
COMPETING INTEREST:
The authors declared no conflict of interest.
AUTHORS’ CONTRIBUTION:
AB: Conceived and designed, data collection, manuscript drafting and editing.
YHK: Critical reviewing and editing.
MZB: Data analysis and correspondence.
All authors approved the final version of the manuscript to be published.
REFERENCES